Paying Attention to SQuAD: Exploring Bidirectional Attention Flow

نویسندگان

  • Heather Blundell
  • Lucy Li
چکیده

With the goal of automated reading comprehension, we apply a neural network with Bidirectional Attention Flow (BiDAF) to the Stanford Question Answering Dataset (SQuAD) and achieve F1 and Exact Match (EM) scores close to the original paper with a single model. We obtain a test F1 score of 76.037 and test EM score of 66.663. Our model includes Character-level CNN embeddings, a Highway Network layer, a Phrase Embedding layer, a Modeling layer, and smart span selection. We also explored expanding the model with feature engineering and an Answer Pointer output layer, which did not further improve our best model. We analyze our model’s performance across categories of contexts, questions, and answers, and compare baseline attention with BiDAF.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Bidirectional Attention Flow for Machine Comprehension

Machine comprehension (MC), answering a query about a given context paragraph, requires modeling complex interactions between the context and the query. Recently, attention mechanisms have been successfully extended to MC. Typically these methods use attention to focus on a small portion of the context and summarize it with a fixed-size vector, couple attentions temporally, and/or often form a ...

متن کامل

Attention-based Recurrent Neural Networks for Question Answering

Machine Comprehension (MC) of text is an important problem in Natural Language Processing (NLP) research, and the task of Question Answering (QA) is a major way of assessing MC outcomes. One QA dataset that has gained immense popularity recently is the Stanford Question Answering Dataset (SQuAD). Successful models for SQuAD have all involved the use of Recurrent Neural Network (RNN), and most o...

متن کامل

CS224n Assignment 4: Machine Comprehension with Exploration on Attention Mechanism

This goal of this paper is to perform the prediction task on SQuAD dataset about reading comprehension. Given a pair of context paragraph and a question, we’ll output an answer. To do this, a model is built combining the idea of Bidirectional LSTM and attention flow mechanism. The basic architecture and setup details of the model are introduced, so do the summary of performance and error analys...

متن کامل

Seq2seq-Attention Question Answering Model

A sequence-to-sequence attention reading comprehension model was implemented to fulfill Question Answering task defined in Stanford Question Answering Dataset (SQuAD). The basic structure was bidirectional LSTM (BiLSTM) encodings with attention mechanism as well as BiLSTM decoding. Several adjustments such as dropout, learning rate decay, and gradients clipping were used. Finally, the model ach...

متن کامل

Looking for Leptonic Cp Violation with Neutrinos

I discuss some theoretical aspects of how to observe leptonic CP violation. It is divided into two parts, one for CP violation due to Majorana, and the other more conventional leptonic Kobayashi-Maskawa (KM) phases. In the first part, I estimate the effect of Majorana phase to observable of neutrinoless double beta decay experiments by paying a careful attention to the definition of the atmosph...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2018